Fast and Scalable Structural SVM with Slack Rescaling
نویسندگان
چکیده
We present an efficient method for training slackrescaled structural SVM. Although finding the most violating label in a margin-rescaled formulation is often easy since the target function decomposes with respect to the structure, this is not the case for a slack-rescaled formulation, and finding the most violated label might be very difficult. Our core contribution is an efficient method for finding the most-violatinglabel in a slack-rescaled formulation, given an oracle that returns the most-violating-label in a (slightly modified) margin-rescaled formulation. We show that our method enables accurate and scalable training for slack-rescaled SVMs, reducing runtime by an order of magnitude compared to previous approaches to slack-rescaled SVMs.
منابع مشابه
Slack and Margin Rescaling as Convex Extensions of Supermodular Functions
Slack and margin rescaling are variants of the structured output SVM. They define convex surrogates to task specific loss functions, which, when specialized to nonadditive loss functions for multi-label problems, yield extensions to increasing set functions. We demonstrate in this paper that we may use these concepts to define polynomial time convex extensions of arbitrary supermodular function...
متن کاملSupplementary Material for “DivMCuts: Faster Training of Structural SVMs with Diverse M-Best Cutting-Planes”
This section reviews the notation used in the main paper and revisits cutting-plane methods for training structured-output predictors. Notation. For any positive integer n we use [n] as shorthand for the set {1, 2, . . . , n}. We use y for a structured-output, and Y = (y1, . . . ,y|Y|) for a tuple of structured-outputs. Given a training dataset of input-output pairs {(xi,yi) |xi ∈X ,yi ∈Y}, we ...
متن کاملA Parallel Way to Select the Parameters of SVM Based on the Ant Optimization Algorithm
A large number of experimental data shows that Support Vector Machine (SVM) algorithm has obvious a large advantages in text classification, handwriting recognition, image classification, bioinformatics, and some other fields. To some degree, the optimization of SVM depends on its kernel function and Slack variable, the determinant of which is its parameters δ and c in the classification functi...
متن کاملA Note on the Generalization Performance of Kernel Classifiers with Margin
We present distribution independent bounds on the generalization misclassification performance of a family of kernel classifiers with margin. Support Vector Machine classifiers (SVM) stem out of this class of machines. The bounds are derived through computations of the Vγ dimension of a family of loss functions where the SVM one belongs to. Bounds that use functions of margin distributions (i.e...
متن کاملThe Benefits of Modeling Slack Variables in SVMs
In this letter, we explore the idea of modeling slack variables in support vector machine (SVM) approaches. The study is motivated by SVM+, which models the slacks through a smooth correcting function that is determined by additional (privileged) information about the training examples not available in the test phase. We take a closer look at the meaning and consequences of smooth modeling of s...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2016